Magnitude and Similarity Based Variable Rate Filter Pruning for Efficient Convolution Neural Networks
نویسندگان
چکیده
The superior performance of the recent deep learning models comes at cost a significant increase in computational complexity, memory use, and power consumption. Filter pruning is one effective neural network compression techniques suitable for model deployment modern low-power edge devices. In this paper, we propose loss-aware filter Magnitude Similarity based Variable rate Pruning (MSVFP) technique. We studied several selection criteria on magnitude similarity among filters within convolution layer, assumption that sensitivity each layer throughout different, unlike conventional fixed methods, our algorithm using automatically finds network. addition, proposed adapts two different to remove weak as well redundant score respectively. Finally, iterative retraining approach are used maintain accuracy during its target float point operations (FLOPs) reduction rate. algorithm, small number steps sufficient prevent an abrupt drop Experiments with commonly VGGNet ResNet CIFAR-10 ImageNet benchmark show superiority method over existing methods literature. Notably, VGG-16, ResNet-56, ResNet-110 dataset even improved original more than 50% FLOPs. Additionally, ResNet-50 reduces FLOPs by 42% negligible accuracy.
منابع مشابه
Pruning Convolutional Neural Networks for Resource Efficient Transfer Learning
We propose a new framework for pruning convolutional kernels in neural networks to enable efficient inference, focusing on transfer learning where large and potentially unwieldy pretrained networks are adapted to specialized tasks. We interleave greedy criteria-based pruning with fine-tuning by backpropagation—a computationally efficient procedure that maintains good generalization in the prune...
متن کاملPruning Convolutional Neural Networks for Resource Efficient Inference
We propose a new formulation for pruning convolutional kernels in neural networks to enable efficient inference. We interleave greedy criteria-based pruning with finetuning by backpropagation—a computationally efficient procedure that maintains good generalization in the pruned network. We propose a new criterion based on Taylor expansion that approximates the change in the cost function induce...
متن کاملEfficient Pruning Method for Ensemble Self-Generating Neural Networks
Recently, multiple classifier systems (MCS) have been used for practical applications to improve classification accuracy. Self-generating neural networks (SGNN) are one of the suitable base-classifiers for MCS because of their simple setting and fast learning. However, the computation cost of the MCS increases in proportion to the number of SGNN. In this paper, we propose an efficient pruning m...
متن کاملSTDP Based Pruning of Connections and Weight Quantization in Spiking Neural Networks for Energy Efficient Recognition
Spiking Neural Networks (SNNs) with a large number of weights and varied weight distribution can be difficult to implement in emerging in-memory computing hardware due to the limitations on crossbar size (implementing dot product), the constrained number of conductance levels in non-CMOS devices and the power budget. We present a sparse SNN topology where non-critical connections are pruned to ...
متن کاملSimilarity-based Heterogeneous Neural Networks
This research introduces a general class of functions serving as generalized neuron models to be used in artificial neural networks. They are cast in the common framework of computing a similarity function, a flexible definition of a neuron as a pattern recognizer. The similarity endows the model with a clear conceptual view and leads naturally to handle heterogeneous information, in the form o...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Applied sciences
سال: 2022
ISSN: ['2076-3417']
DOI: https://doi.org/10.3390/app13010316